13. MDPs, Part 1

MDPs

Over the next several videos, you'll learn all about how to rigorously define a reinforcement learning problem as a Markov Decision Process (MDP) .

Towards this goal, we'll begin with an example!

MDPs, Part 1

## Notes

In general, the state space \mathcal{S} is the set of all nonterminal states .

In continuing tasks (like the recycling task detailed in the video), this is equivalent to the set of all states .

In episodic tasks, we use \mathcal{S}^+ to refer to the set of all states, including terminal states .

The action space \mathcal{A} is the set of possible actions available to the agent.

In the event that there are some states where only a subset of the actions are available, we can also use \mathcal{A}(s) to refer to the set of actions available in state s\in\mathcal{S} .